episode difficulty
A Experimental setup
A.1 Datasets We use two standardized few-shot image classification datasets. We also use the test splits of the following four datasets, as defined by Triantafillou et al. [57]. CUB-200: CUB-200 was collected by Welinder et al. The test split contains 30 classes. A.2 Network architectures We train two of the most popular network architectures in few-shot learning literature. Episode difficulty is approximately normally distributed - density plots.
Uniform Sampling over Episode Difficulty Sébastien M. R. Arnold
Building on this method, we perform an extensive analysis and find that sampling uniformly over episode difficulty outperforms other sampling schemes, including curriculum and easy-/hard-mining. As the proposed sampling method is algorithm agnostic, we can leverage these insights to improve few-shot learning accuracies across many episodic training algorithms.
- North America > United States > California > Los Angeles County > Los Angeles (0.14)
- North America > United States > Massachusetts > Plymouth County > Norwell (0.04)
- North America > United States > Illinois > Cook County > Chicago (0.04)
- (2 more...)
- Media > Television (0.46)
- Leisure & Entertainment (0.46)
- North America > United States > California > Los Angeles County > Los Angeles (0.14)
- North America > United States > Massachusetts > Plymouth County > Norwell (0.04)
- North America > United States > Illinois > Cook County > Chicago (0.04)
- (2 more...)
- Media > Television (0.46)
- Leisure & Entertainment (0.46)
Uniform Sampling over Episode Difficulty
Episodic training is a core ingredient of few-shot learning to train models on tasks with limited labelled data. Despite its success, episodic training remains largely understudied, prompting us to ask the question: what is the best way to sample episodes? In this paper, we first propose a method to approximate episode sampling distributions based on their difficulty. Building on this method, we perform an extensive analysis and find that sampling uniformly over episode difficulty outperforms other sampling schemes, including curriculum and easy-/hard-mining. As the proposed sampling method is algorithm agnostic, we can leverage these insights to improve few-shot learning accuracies across many episodic training algorithms.
Uniform Sampling over Episode Difficulty
Arnold, Sébastien M. R., Dhillon, Guneet S., Ravichandran, Avinash, Soatto, Stefano
Episodic training is a core ingredient of few-shot learning to train models on tasks with limited labelled data. Despite its success, episodic training remains largely understudied, prompting us to ask the question: what is the best way to sample episodes? In this paper, we first propose a method to approximate episode sampling distributions based on their difficulty. Building on this method, we perform an extensive analysis and find that sampling uniformly over episode difficulty outperforms other sampling schemes, including curriculum and easy-/hard-mining. As the proposed sampling method is algorithm agnostic, we can leverage these insights to improve few-shot learning accuracies across many episodic training algorithms. We demonstrate the efficacy of our method across popular few-shot learning datasets, algorithms, network architectures, and protocols.
- North America > United States > New York > New York County > New York City (0.14)
- North America > United States > California > Los Angeles County > Los Angeles (0.14)
- North America > United States > Massachusetts > Plymouth County > Norwell (0.04)
- (2 more...)
- Media > Television (0.46)
- Leisure & Entertainment (0.46)